51 research outputs found
Act-Aware Slot-Value Predicting in Multi-Domain Dialogue State Tracking
As an essential component in task-oriented dialogue systems, dialogue state
tracking (DST) aims to track human-machine interactions and generate state
representations for managing the dialogue. Representations of dialogue states
are dependent on the domain ontology and the user's goals. In several
task-oriented dialogues with a limited scope of objectives, dialogue states can
be represented as a set of slot-value pairs. As the capabilities of dialogue
systems expand to support increasing naturalness in communication,
incorporating dialogue act processing into dialogue model design becomes
essential. The lack of such consideration limits the scalability of dialogue
state tracking models for dialogues having specific objectives and ontology. To
address this issue, we formulate and incorporate dialogue acts, and leverage
recent advances in machine reading comprehension to predict both categorical
and non-categorical types of slots for multi-domain dialogue state tracking.
Experimental results show that our models can improve the overall accuracy of
dialogue state tracking on the MultiWOZ 2.1 dataset, and demonstrate that
incorporating dialogue acts can guide dialogue state design for future
task-oriented dialogue systems.Comment: Published in Spoken Dialogue Systems I, Interspeech 2021. Code is now
publicly available on Github: https://github.com/youlandasu/ACT-AWARE-DS
Schema Graph-Guided Prompt for Multi-Domain Dialogue State Tracking
Tracking dialogue states is an essential topic in task-oriented dialogue
systems, which involve filling in the necessary information in pre-defined
slots corresponding to a schema. While general pre-trained language models have
been shown effective in slot-filling, their performance is limited when applied
to specific domains. We propose a graph-based framework that learns
domain-specific prompts by incorporating the dialogue schema. Specifically, we
embed domain-specific schema encoded by a graph neural network into the
pre-trained language model, which allows for relations in the schema to guide
the model for better adaptation to the specific domain. Our experiments
demonstrate that the proposed graph-based method outperforms other multi-domain
DST approaches while using similar or fewer trainable parameters. We also
conduct a comprehensive study of schema graph architectures, parameter usage,
and module ablation that demonstrate the effectiveness of our model on
multi-domain dialogue state tracking
- …